1,105,851 research outputs found
Deep learning for detection and segmentation of artefact and disease instances in gastrointestinal endoscopy
The Endoscopy Computer Vision Challenge (EndoCV) is a crowd-sourcing initiative to address eminent problems in developing
reliable computer aided detection and diagnosis endoscopy systems and suggest a pathway for clinical translation
of technologies. Whilst endoscopy is a widely used diagnostic and treatment tool for hollow-organs, there are several core
challenges often faced by endoscopists, mainly: 1) presence of multi-class artefacts that hinder their visual interpretation, and
2) difficulty in identifying subtle precancerous precursors and cancer abnormalities. Artefacts often affect the robustness of
deep learning methods applied to the gastrointestinal tract organs as they can be confused with tissue of interest. EndoCV2020
challenges are designed to address research questions in these remits. In this paper, we present a summary of methods
developed by the top 17 teams and provide an objective comparison of state-of-the-art methods and methods designed by
the participants for two sub-challenges: i) artefact detection and segmentation (EAD2020), and ii) disease detection and
segmentation (EDD2020). Multi-center, multi-organ, multi-class, and multi-modal clinical endoscopy datasets were compiled
for both EAD2020 and EDD2020 sub-challenges. The out-of-sample generalization ability of detection algorithms was also
evaluated. Whilst most teams focused on accuracy improvements, only a few methods hold credibility for clinical usability. The
best performing teams provided solutions to tackle class imbalance, and variabilities in size, origin, modality and occurrences
by exploring data augmentation, data fusion, and optimal class thresholding techniques
Recommended from our members
Medical Image Data and Datasets in the Era of Machine Learning-Whitepaper from the 2016 C-MIMI Meeting Dataset Session.
At the first annual Conference on Machine Intelligence in Medical Imaging (C-MIMI), held in September 2016, a conference session on medical image data and datasets for machine learning identified multiple issues. The common theme from attendees was that everyone participating in medical image evaluation with machine learning is data starved. There is an urgent need to find better ways to collect, annotate, and reuse medical imaging data. Unique domain issues with medical image datasets require further study, development, and dissemination of best practices and standards, and a coordinated effort among medical imaging domain experts, medical imaging informaticists, government and industry data scientists, and interested commercial, academic, and government entities. High-level attributes of reusable medical image datasets suitable to train, test, validate, verify, and regulate ML products should be better described. NIH and other government agencies should promote and, where applicable, enforce, access to medical image datasets. We should improve communication among medical imaging domain experts, medical imaging informaticists, academic clinical and basic science researchers, government and industry data scientists, and interested commercial entities
DRINet for medical image segmentation
Convolutional neural networks (CNNs) have revolutionized medical image analysis over the past few years. The UNet architecture is one of the most well-known CNN architectures for semantic segmentation and has achieved remarkable successes in many different medical image segmentation applications. The U-Net architecture consists of standard convolution layers, pooling layers, and upsampling layers. These convolution layers learn representative features of input images and construct segmentations based on the features. However, the features learned by standard convolution layers are not distinctive when the differences among different categories are subtle in terms of intensity, location, shape, and size. In this paper, we propose a novel CNN architecture, called Dense-Res-Inception Net (DRINet), which addresses this challenging problem. The proposed DRINet consists of three blocks, namely a convolutional block with dense connections, a deconvolutional block with residual Inception modules, and an unpooling block. Our proposed architecture outperforms the U-Net in three different challenging applications, namely multi-class segmentation of cerebrospinal fluid (CSF) on brain CT images, multi-organ segmentation on abdominal CT images, multi-class brain tumour segmentation on MR images
- …